Smaller generalization error derived for a deep residual neural network compared with shallow networks

نویسندگان

چکیده

Abstract Estimates of the generalization error are proved for a residual neural network with $L$ random Fourier features layers $\bar z_{\ell +1}=\bar z_\ell + \textrm {Re}\sum _{k=1}^K\bar b_{\ell k}\,e^{\textrm {i}\omega _{\ell k}\bar }+ c_{\ell ^{\prime}_{\ell k}\cdot x}$. An optimal distribution frequencies $(\omega k},\omega k})$ $e^{\textrm }$ and x}$ is derived. This derivation based on corresponding approximation function values $f(x)$. The turns out to be smaller than estimate ${\|\hat f\|^2_{L^1({\mathbb {R}}^d)}}/{(KL)}$ features, one hidden layer same total number nodes $KL$, in case $L^\infty $-norm $f$ much less $L^1$-norm its transform $\hat f$. understanding an used construct new training method deep network. Promising performance proposed algorithm demonstrated computational experiments.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Inferring Functional Neural Connectivity with Deep Residual Convolutional Networks

Measuring synaptic connectivity in large neuronal populations remains a major goal of modern neuroscience. While this connectivity is traditionally revealed by anatomical methods such as electron microscopy, an efficient alternative is to computationally infer functional connectivity from recordings of neural activity. However, these statistical techniques still require further refinement befor...

متن کامل

GoGoGo: Improving Deep Neural Network Based Go Playing AI with Residual Networks

The game of Go has a long history and has been viewed as the most challanging classical game due to the enormous amount of possible moves and the lack of precise evaulation tools. AlphaGo [13], a Go-playing AI built by Google DeepMind, used a new approach of combining deep neural networks with tree search to solve the Go playing problem. It narrows down the search space by introducing ”policy n...

متن کامل

Reversible Architectures for Arbitrarily Deep Residual Neural Networks

Recently, deep residual networks have been successfully applied in many computer vision and natural language processing tasks, pushing the state-of-the-art performance with deeper and wider architectures. In this work, we interpret deep residual networks as ordinary differential equations (ODEs), which have long been studied in mathematics and physics with rich theoretical and empirical success...

متن کامل

On the complexity of shallow and deep neural network classifiers

Recently, deep networks were proved to be more effective than shallow architectures to face complex real–world applications. However, theoretical results supporting this claim are still few and incomplete. In this paper, we propose a new topological measure to study how the depth of feedforward networks impacts on their ability of implementing high complexity functions. Upper and lower bounds o...

متن کامل

Generalization Tower Network: A Novel Deep Neural Network Architecture for Multi-Task Learning

Deep learning (DL) advances state-of-the-art reinforcement learning (RL), by incorporating deep neural networks in learning representations from the input to RL. However, the conventional deep neural network architecture is limited in learning representations for multi-task RL (MT-RL), as multiple tasks can refer to different kinds of representations. In this paper, we thus propose a novel deep...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Ima Journal of Numerical Analysis

سال: 2022

ISSN: ['1464-3642', '0272-4979']

DOI: https://doi.org/10.1093/imanum/drac049